45 research outputs found
Online Out-of-Domain Detection for Automated Driving
Ensuring safety in automated driving is a major challenge for the automotive
industry. Special attention is paid to artificial intelligence, in particular
to Deep Neural Networks (DNNs), which is considered a key technology in the
realization of highly automated driving. DNNs learn from training data, which
means that they only achieve good accuracy within the underlying data
distribution of the training data. When leaving the training domain, a
distributional shift is caused, which can lead to a drastic reduction of
accuracy. In this work, we present a proof of concept for a safety mechanism
that can detect the leaving of the domain online, i.e. at runtime. In our
experiments with the Synthia data set we can show that a 100 % correct
detection of whether the input data is inside or outside the domain is
achieved. The ability to detect when the vehicle leaves the domain can be an
important requirement for certification.Comment: Machine Learning in Certified Systems (MLCS) Workshop, 14.-15.01.202
Efficient Multi-Task RGB-D Scene Analysis for Indoor Environments
Semantic scene understanding is essential for mobile agents acting in various
environments. Although semantic segmentation already provides a lot of
information, details about individual objects as well as the general scene are
missing but required for many real-world applications. However, solving
multiple tasks separately is expensive and cannot be accomplished in real time
given limited computing and battery capabilities on a mobile platform. In this
paper, we propose an efficient multi-task approach for RGB-D scene
analysis~(EMSANet) that simultaneously performs semantic and instance
segmentation~(panoptic segmentation), instance orientation estimation, and
scene classification. We show that all tasks can be accomplished using a single
neural network in real time on a mobile platform without diminishing
performance - by contrast, the individual tasks are able to benefit from each
other. In order to evaluate our multi-task approach, we extend the annotations
of the common RGB-D indoor datasets NYUv2 and SUNRGB-D for instance
segmentation and orientation estimation. To the best of our knowledge, we are
the first to provide results in such a comprehensive multi-task setting for
indoor scene analysis on NYUv2 and SUNRGB-D.Comment: To be published in IEEE International Joint Conference on Neural
Networks (IJCNN) 202
Point cloud hand-object segmentation using multimodal imaging with thermal and color data for safe robotic object handover
This paper presents an application of neural networks operating on multimodal 3D data (3D point cloud, RGB, thermal) to effectively and precisely segment human hands and objects held in hand to realize a safe human–robot object handover. We discuss the problems encountered in building a multimodal sensor system, while the focus is on the calibration and alignment of a set of cameras including RGB, thermal, and NIR cameras. We propose the use of a copper–plastic chessboard calibration target with an internal active light source (near-infrared and visible light). By brief heating, the calibration target could be simultaneously and legibly captured by all cameras. Based on the multimodal dataset captured by our sensor system, PointNet, PointNet++, and RandLA-Net are utilized to verify the effectiveness of applying multimodal point cloud data for hand–object segmentation. These networks were trained on various data modes (XYZ, XYZ-T, XYZ-RGB, and XYZ-RGB-T). The experimental results show a significant improvement in the segmentation performance of XYZ-RGB-T (mean Intersection over Union: 82.8% by RandLA-Net) compared with the other three modes (77.3% by XYZ-RGB, 35.7% by XYZ-T, 35.7% by XYZ), in which it is worth mentioning that the Intersection over Union for the single class of hand achieves 92.6%
Approaching a person in a socially acceptable manner using expanding random trees
In real world scenarios for mobile robots, socially acceptable navigation is a key component to interact naturally with other persons. On the one hand this enables a robot to behave more human-like, and on the other hand it increases the acceptance of the user towards the robot as an interaction partner. As part of this research field, we present in this paper a strategy of approaching a person in a socially acceptable manner. Therefore, we use the theory of ”personal space” and present a method of modeling this space to enable a mobile robot to approach a person from the front. We use a standard Dynamic Window Approach to control the robot motion and, since the personal space model could not be used directly, a graph planner in configuration space, to plan an optimal path by expanding the graph with the use of the DWA’s update rule. Additionally, we give a proof of concept with first preliminary experiments